Spectral Convolution Networks

نویسندگان

  • Maria Francesca
  • Arthur Hughes
  • David Gregg
چکیده

Previous research has shown that computation of convolution in the frequency domain provides a significant speedup versus traditional convolution network implementations. However, this performance increase comes at the expense of repeatedly computing the transform and its inverse in order to apply other network operations such as activation, pooling, and dropout. We show, mathematically, how convolution and activation can both be implemented in the frequency domain using either the Fourier or Laplace transformation. The main contributions are a description of spectral activation under the Fourier transform and a further description of an efficient algorithm for computing both convolution and activation under the Laplace transform. By computing both the convolution and activation functions in the frequency domain, we can reduce the number of transforms required, as well as reducing overall complexity. Our description of a spectral activation function, together with existing spectral analogs of other network functions may then be used to compose a fully spectral implementation of a convolution network. 1 Motivation Convolution networks are used for machine learning problems such as image classification, natural language processing, and recommendation systems [4, 3, 8]. They are represented as a graph of operators which are typically sequentially applied to some input image, eventually yielding a classification for that input. Convolution is an expensive operation which is replicated repeatedly within a single network. Computation of convolution in the frequency domain under the Fourier transform has been shown to provide a significant speedup versus traditional convolution network implementations [9, 4]. However, typically activation is run following convolution in practise, and previous researchers have been unable to find a spectral implementation for both. In this paper we describe spectral representations of the max activation function a(x) paying particular attention to computational complexity. Specifically, a(x) = max(0, x). The abbreviation ReLU is used throughout to designate the part of the network which computes this function.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Numerical resolution of large deflections in cantilever beams by Bernstein spectral method and a convolution quadrature.

The mathematical modeling of the large deflections for the cantilever beams leads to a nonlinear differential equation with the mixed boundary conditions. Different numerical methods have been implemented by various authors for such problems. In this paper, two novel numerical techniques are investigated for the numerical simulation of the problem. The first is based on a spectral method utiliz...

متن کامل

Some Results on Convex Spectral Functions: I

In this paper, we give a fundamental convexity preserving for spectral functions. Indeed, we investigate infimal convolution, Moreau envelope and proximal average for convex spectral functions, and show that this properties are inherited from the properties of its corresponding convex function. This results have many applications in Applied Mathematics such as semi-definite programmings and eng...

متن کامل

Wavelet Convolutional Neural Networks for Texture Classification

Texture classification is an important and challenging problem in many image processing applications. While convolutional neural networks (CNNs) achieved significant successes for image classification, texture classification remains a difficult problem since textures usually do not contain enough information regarding the shape of object. In image processing, texture classification has been tra...

متن کامل

Bipolar spectral associative memories

Nonlinear spectral associative memories are proposed as quantized frequency domain formulations of nonlinear, recurrent associative memories in which volatile network attractors are instantiated by attractor waves. In contrast to conventional associative memories, attractors encoded in the frequency domain by convolution may be viewed as volatile online inputs, rather than nonvolatile, off-line...

متن کامل

Inferring Phonemic Classes from CNN Activation Maps Using Clustering Techniques

Today’s state-of-art in speech recognition involves deep neural networks (DNN). These last years, a certain research effort has been invested in characterizing the feature representations learned by DNNs. In this paper, we focus on convolutional neural networks (CNN) trained for phoneme recognition in French. We report clustering experiments performed on activation maps extracted from the diffe...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1611.05378  شماره 

صفحات  -

تاریخ انتشار 2016